Improving matrix-vector multiplication via lossless grammar-compressed matrices
نویسندگان
چکیده
As nowadays Machine Learning (ML) techniques are generating huge data collections, the problem of how to efficiently engineer their storage and operations is becoming paramount importance. In this article we propose a new lossless compression scheme for real-valued matrices which achieves efficient performance in terms ratio time linear-algebra operations. Experiments show that, as compressor, our tool clearly superior gzip it usually within 20% xz ratio. addition, compressed format supports matrix-vector multiplications space proportional size representation, unlike that require full decompression matrix. To knowledge compressor first one achieving complexities match theoretical limit expressed by k -th order statistical entropy input. achieve further time/space reductions, column-reordering algorithms hinging on novel column-similarity score. Our experiments various sets ML column reordering can yield reduction up 16% peak memory usage during multiplication. Finally, compare proposal against state-of-the-art Compressed Linear Algebra (CLA) approach showing ours runs always at least twice faster (in multi-thread setting), better occupancy usage. This experimentally confirms provably effective bounds compressed-matrix approach.
منابع مشابه
Matrix-Vector Multiplication via Erasure Decoding
The problem of fast evaluation of a matrix-vector product over GF (2) is considered. The problem is reduced to erasure decoding of a linear error-correcting code. A large set of redundant parity check equations for this code is generated. The multiplication algorithm is derived by tracking the execution of the message-passing algorithm on the obtained set of parity check equations. The obtained...
متن کاملCompressed Modular Matrix Multiplication
Matrices of integers modulo a small prime can be compressed by storing several entries into a single machine word. Modular addition is performed by addition and possibly subtraction of a word containing several times the modulus. We show how modular multiplication can also be performed. In terms of arithmetic operations, the gain over classical matrix multiplication is equal to the number of in...
متن کاملDecomposing Irregularly Sparse Matrices for Parallel Matrix-Vector Multiplication
In this work we show the de ciencies of the graph model for decomposing sparse matrices for parallel matrix vector multiplica tion Then we propose two hypergraph models which avoid all de cien cies of the graph model The proposed models reduce the decomposition problem to the well known hypergraph partitioning problem widely en countered in circuit partitioning in VLSI We have implemented fast ...
متن کاملImproving Memory-System Performance of Sparse Matrix-Vector Multiplication
Sparse matrix-vector multiplication is an important kernel that often runs inefficiently on superscalar RISC processors. This paper describes techniques that increase instruction-level parallelism and improve performance. The techniques include reordering to reduce cache misses originally due to Das et al., blocking to reduce load instructions, and prefetching to prevent multiple load-store uni...
متن کاملOn improving the performance of sparse matrix-vector multiplication
We analyze single-node performance of sparse matrix-vector multiplication by investigating issues of data locality and ne-grained parallelism. We examine the data-locality characteristics of the compressed-sparse-row representation and consider improvements in locality through matrix permutation. Motivated by potential improvements in ne-grained parallelism, we evaluate modiied sparse-matrix re...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Proceedings of the VLDB Endowment
سال: 2022
ISSN: ['2150-8097']
DOI: https://doi.org/10.14778/3547305.3547321